Equivalence of Entropy Regularization and Relative-Entropy Proximal Method
نویسنده
چکیده
We consider two entropy-based interior point methods that solve LP relaxations of MAP estimation in graphical models: (1) an entropy-regularization method and (2) a relative-entropy proximal method. Using the fact that relative-entropy is the Bregman distance induced by entropy, we show that the two approaches are actually equivalent. The purpose of this note is to show one connection between the two approaches described in [1, 2]. Another connection between these two works is that both use distributed iterative-scaling/Bregman-projections algorithms [3, 4] to solve the “inner-loop” optimizations required by the methods summarized below. This second connection, however, is not explored in this present note. Introduction For the sake of this note, we consider the exponential family of probability distributions on n binary variables x = (x1, . . . , xn) ∈ {0, 1} : P (x) = exp{θφ(x) − Φ(θ)} = 1 Z(θ) exp{ ∑
منابع مشابه
Observational Modeling of the Kolmogorov-Sinai Entropy
In this paper, Kolmogorov-Sinai entropy is studied using mathematical modeling of an observer $ Theta $. The relative entropy of a sub-$ sigma_Theta $-algebra having finite atoms is defined and then the ergodic properties of relative semi-dynamical systems are investigated. Also, a relative version of Kolmogorov-Sinai theorem is given. Finally, it is proved that the relative entropy of a...
متن کاملSome properties of the parametric relative operator entropy
The notion of entropy was introduced by Clausius in 1850, and some of the main steps towards the consolidation of the concept were taken by Boltzmann and Gibbs. Since then several extensions and reformulations have been developed in various disciplines with motivations and applications in different subjects, such as statistical mechanics, information theory, and dynamical systems. Fujii and Kam...
متن کاملA short variational proof of equivalence between policy gradients and soft Q learning
Two main families of reinforcement learning algorithms, Q-learning and policy gradients, have recently been proven to be equivalent when using a softmax relaxation on one part, and an entropic regularization on the other. We relate this result to the well-known convex duality of Shannon entropy and the softmax function. Such a result is also known as the Donsker-Varadhan formula. This provides ...
متن کاملRelative Entropy Rate between a Markov Chain and Its Corresponding Hidden Markov Chain
In this paper we study the relative entropy rate between a homogeneous Markov chain and a hidden Markov chain defined by observing the output of a discrete stochastic channel whose input is the finite state space homogeneous stationary Markov chain. For this purpose, we obtain the relative entropy between two finite subsequences of above mentioned chains with the help of the definition of...
متن کاملA note on inequalities for Tsallis relative operator entropy
In this short note, we present some inequalities for relative operator entropy which are generalizations of some results obtained by Zou [Operator inequalities associated with Tsallis relative operator entropy, {em Math. Inequal. Appl.} {18} (2015), no. 2, 401--406]. Meanwhile, we also show some new lower and upper bounds for relative operator entropy and Tsallis relative o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008